Orders Of Magnitude (computing)
   HOME

TheInfoList



OR:

This list compares various amounts of computing power in instructions per second organized by
order of magnitude An order of magnitude is an approximation of the logarithm of a value relative to some contextually understood reference value, usually 10, interpreted as the base of the logarithm and the representative of values of magnitude one. Logarithmic dis ...
in
FLOPS In computing, floating point operations per second (FLOPS, flops or flop/s) is a measure of computer performance, useful in fields of scientific computations that require floating-point calculations. For such cases, it is a more accurate meas ...
.
Scientific E notation index: 2 , 3 , 6 , 9 , 12 , 15 , 18 , 21 , 24 , >24
__TOC__


Deciscale computing (10−1)

* 5×10−1: Computing power of the average human mental calculation for multiplication using pen and paper


Scale computing (100)

* 1 OP/S: Power of an average human performing calculations using pen and paper * 1 OP/S: Computing power of
Zuse Z1 The Z1 was a motor-driven mechanical computer designed by Konrad Zuse from 1936 to 1937, which he built in his parents' home from 1936 to 1938. It was a binary electrically driven mechanical calculator with limited programmability, reading in ...
* 5 OP/S: World record for addition set


Decascale computing (101)

* 5×101: Upper end of serialized human perception computation (light bulbs do not flicker to the human observer)


Hectoscale computing (102)

* 2.2×102: Upper end of serialized human throughput. This is roughly expressed by the lower limit of accurate event placement on small scales of time (The swing of a conductor's arm, the reaction time to lights on a drag strip, etc.) * 2×102:
IBM 602 The IBM 602 Calculating Punch, introduced in 1946, was an electromechanical calculator capable of addition, subtraction, multiplication, and division. The 602 was IBM's first machine that did division. (The IBM 601, introduced in 1931, only multip ...
1946 computer.


Kiloscale computing (103)

* 92×103:
Intel 4004 The Intel 4004 is a 4-bit central processing unit (CPU) released by Intel Corporation in 1971. Sold for US$60, it was the first commercially produced microprocessor, and the first in a long line of Intel CPUs. The 4004 was the first signific ...
First commercially available full function CPU on a chip, released in 1971 * 500×103
Colossus computer Colossus was a set of computers developed by British codebreakers in the years 1943–1945 to help in the cryptanalysis of the Lorenz cipher. Colossus used thermionic valves (vacuum tubes) to perform Boolean and counting operations. Colossus ...
vacuum tube A vacuum tube, electron tube, valve (British usage), or tube (North America), is a device that controls electric current flow in a high vacuum between electrodes to which an electric voltage, potential difference has been applied. The type kn ...
supercomputer 1943


Megascale computing (106)

* 1×106: Computing power of the
Motorola 68000 The Motorola 68000 (sometimes shortened to Motorola 68k or m68k and usually pronounced "sixty-eight-thousand") is a 16/32-bit complex instruction set computer (CISC) microprocessor, introduced in 1979 by Motorola Semiconductor Products Sector ...
commercial computer introduced in 1979. This is also the minimum computing power of a Type 0 Kardashev civilization. * 1.2×106: IBM 7030 "Stretch" transistorized supercomputer 1961


Gigascale computing (109)

* 1×109:
ILLIAC IV The ILLIAC IV was the first massively parallel computer. The system was originally designed to have 256 64-bit floating point units (FPUs) and four central processing units (CPUs) able to process 1 billion operations per second. Due to budget con ...
1972 supercomputer does first
computational fluid dynamics Computational fluid dynamics (CFD) is a branch of fluid mechanics that uses numerical analysis and data structures to analyze and solve problems that involve fluid flows. Computers are used to perform the calculations required to simulate th ...
problems * 1.354×109:
Intel Pentium III The Pentium III (marketed as Intel Pentium III Processor, informally PIII or P3) brand refers to Intel's 32-bit x86 desktop and mobile CPUs based on the sixth-generation P6 microarchitecture introduced on February 28, 1999. The brand's initial p ...
commercial computing 1999 * 147.6×109: Intel Core i7-980X Extreme Edition commercial computing 2010


Terascale computing (1012)

* 1.34×1012:
Intel ASCI Red ASCI Red (also known as ASCI Option Red or TFLOPS) was the first computer built under the Accelerated Strategic Computing Initiative ( ASCI), the supercomputing initiative of the United States government created to help the maintenance of the ...
1997 Supercomputer * 1.344×1012 GeForce GTX 480 in 2010 from Nvidia at its peak performance * 4.64×1012: Radeon HD 5970 in 2009 from
AMD Advanced Micro Devices, Inc. (AMD) is an American multinational semiconductor company based in Santa Clara, California, that develops computer processors and related technologies for business and consumer markets. While it initially manufactur ...
(under ATI branding) at its peak performance * 5.152×1012: S2050/S2070 1U GPU Computing System from Nvidia * 11.3×1012 : GeForce GTX 1080 Ti in 2017 * 13.7×1012: Radeon RX Vega 64 in 2017 * 15.0×1012: Nvidia Titan V in 2017 * 80×1012:
IBM Watson IBM Watson is a question-answering computer system capable of answering questions posed in natural language, developed in IBM's DeepQA project by a research team led by principal investigator David Ferrucci. Watson was named after IBM's founder ...
* 170×1012:
Nvidia DGX-1 Nvidia DGX is a line of Nvidia-produced servers and workstations which specialize in using GPGPU to accelerate deep learning applications. The typical design of a DGX system is based upon a rackmount chassis with motherboard that carries high per ...
The initial Pascal based DGX-1 delivered 170 teraflops of half precision processing. * 478.2×1012 IBM BlueGene/L 2007 Supercomputer * 960×1012
Nvidia DGX-1 Nvidia DGX is a line of Nvidia-produced servers and workstations which specialize in using GPGPU to accelerate deep learning applications. The typical design of a DGX system is based upon a rackmount chassis with motherboard that carries high per ...
The Volta-based upgrade increased calculation power of
Nvidia DGX-1 Nvidia DGX is a line of Nvidia-produced servers and workstations which specialize in using GPGPU to accelerate deep learning applications. The typical design of a DGX system is based upon a rackmount chassis with motherboard that carries high per ...
to 960
teraflop In computing, floating point operations per second (FLOPS, flops or flop/s) is a measure of computer performance, useful in fields of scientific computations that require floating-point calculations. For such cases, it is a more accurate meas ...
s.


Petascale computing (1015)

* 1.026×1015:
IBM Roadrunner Roadrunner was a supercomputer built by IBM for the Los Alamos National Laboratory in New Mexico, USA. The US$100-million Roadrunner was designed for a peak performance of 1.7 petaflops. It achieved 1.026 petaflops on May 25, 2008, to become the ...
2009 Supercomputer * 1.32×1015: Nvidia GeForce 4000 Series RTX4090 Consumer graphics card, achieves 1.32 petaflops in AI applications, October 2022 * 2×1015:
Nvidia DGX-2 Nvidia DGX is a line of Nvidia-produced servers and workstations which specialize in using GPGPU to accelerate deep learning applications. The typical design of a DGX system is based upon a rackmount chassis with motherboard that carries high per ...
a 2 Petaflop Machine Learning system (the newer DGX A100 has 5 Petaflop performance) *10×1015 (1016): Minimum computing power of a Type I Kardashev civilization * 11.5×1015:
Google Google LLC () is an American multinational technology company focusing on search engine technology, online advertising, cloud computing, computer software, quantum computing, e-commerce, artificial intelligence, and consumer electronics. ...
TPU pod containing 64 second-generation TPUs, May 2017 * 17.17×1015:
IBM Sequoia IBM Sequoia was a Petascale computing, petascale Blue Gene#Blue Gene/Q, Blue Gene/Q supercomputer constructed by IBM for the National Nuclear Security Administration as part of the Advanced Simulation and Computing Program (ASC). It was delivered ...
's LINPACK performance, June 2013 * 20×1015: Roughly the hardware-equivalent of the human brain according to Kurzweil. Published in his 1999 book: The Age of Spiritual Machines: When Computers Exceed Human Intelligence * 33.86×1015:
Tianhe-2 Tianhe-2 or TH-2 (, i.e. 'Milky Way 2') is a 33.86-petaflops supercomputer located in the National Supercomputer Center in Guangzhou, China. It was developed by a team of 1,300 scientists and engineers. It was the world's fastest supercomputer ac ...
's LINPACK performance, June 2013 * 36.8×1015: Estimated computational power required to ''simulate'' a human brain in real time. * 93.01×1015: Sunway TaihuLight's LINPACK performance, June 2016http://top500.org/list/2016/06/ Top500 list, June 2016 *143.5×1015: Summit's LINPACK performance, November 2018


Exascale computing (1018)

* 1×1018: The U.S. Department of Energy and NSA estimated in 2008 that they would need exascale computing around 2018 * 1×1018: Fugaku 2020 supercomputer in single precision mode * 1.1x1018:
Frontier A frontier is the political and geographical area near or beyond a boundary. A frontier can also be referred to as a "front". The term came from French in the 15th century, with the meaning "borderland"—the region of a country that fronts o ...
2022 supercomputer * 1.88×1018: U.S. Summit achieves a peak throughput of this many operations per second, whilst analysing genomic data using a mixture of numerical precisions. * 2.43×1018
Folding@home Folding@home (FAH or F@h) is a volunteer computing project aimed to help scientists develop new therapeutics for a variety of diseases by the means of simulating protein dynamics. This includes the process of protein folding and the movements ...
distributed computing system during
COVID-19 pandemic The COVID-19 pandemic, also known as the coronavirus pandemic, is an ongoing global pandemic of coronavirus disease 2019 (COVID-19) caused by severe acute respiratory syndrome coronavirus 2 (SARS-CoV-2). The novel virus was first identif ...
response


Zettascale computing (1021)

* 1×1021: Accurate global weather estimation on the scale of approximately 2 weeks. Assuming
Moore's law Moore's law is the observation that the number of transistors in a dense integrated circuit (IC) doubles about every two years. Moore's law is an observation and projection of a historical trend. Rather than a law of physics, it is an empir ...
remains applicable, such systems may be feasible around 2035. A zettascale computer system could generate more single floating point data in one second than was stored by any digital means on Earth in the first quarter of 2011.


Beyond zettascale computing (>1021)

*1.12×1036: Estimated computational power of a
Matrioshka brain A matrioshka brain is a hypothetical megastructure of immense computational capacity powered by a Dyson sphere. It was proposed in 1997 by Robert J. Bradbury (1956–2011). It is an example of a class-B stellar engine, employing the entire energy ...
, assuming 1.87×1026
watt The watt (symbol: W) is the unit of power or radiant flux in the International System of Units (SI), equal to 1 joule per second or 1 kg⋅m2⋅s−3. It is used to quantify the rate of energy transfer. The watt is named after James Wa ...
power produced by solar panels and 6 GFLOPS/watt efficiency. *4×1048: Estimated computational power of a Matrioshka brain whose power source is the
Sun The Sun is the star at the center of the Solar System. It is a nearly perfect ball of hot plasma, heated to incandescence by nuclear fusion reactions in its core. The Sun radiates this energy mainly as light, ultraviolet, and infrared radi ...
, the outermost layer operates at 10
kelvin The kelvin, symbol K, is the primary unit of temperature in the International System of Units (SI), used alongside its prefixed forms and the degree Celsius. It is named after the Belfast-born and University of Glasgow-based engineer and phys ...
s, and the constituent parts operate at or near the
Landauer limit Landauer's principle is a physical principle pertaining to the lower theoretical limit of energy consumption of computation. It holds that "any logically irreversible manipulation of information, such as the erasure of a bit or the merging of two ...
and draws power at the efficiency of a
Carnot engine A Carnot heat engine is a heat engine that operates on the Carnot cycle. The basic model for this engine was developed by Nicolas Léonard Sadi Carnot in 1824. The Carnot engine model was graphically expanded by Benoît Paul Émile Clapeyron in ...
*5×1058: Estimated power of a
galaxy A galaxy is a system of stars, stellar remnants, interstellar gas, dust, dark matter, bound together by gravity. The word is derived from the Greek ' (), literally 'milky', a reference to the Milky Way galaxy that contains the Solar System. ...
equivalent in luminosity to the
Milky Way The Milky Way is the galaxy that includes our Solar System, with the name describing the galaxy's appearance from Earth: a hazy band of light seen in the night sky formed from stars that cannot be individually distinguished by the naked eye ...
converted into Matrioshka brains.


See also

*
Futures studies Futures studies, futures research, futurism or futurology is the systematic, interdisciplinary and holistic study of social and technological advancement, and other environmental trends, often for the purpose of exploring how people will li ...
– study of possible, probable, and preferable futures, including making projections of future technological advances *
History of computing hardware (1960s–present) The history of computing hardware starting at 1960 is marked by the conversion from vacuum tube to solid-state devices such as transistors and then integrated circuit (IC) chips. Around 1953 to 1959, discrete transistors started being considered ...
*
List of emerging technologies This is a list of emerging technologies, in-development technical innovations with significant potential in their applications. The criteria for this list is that the technology must: # Exist in some way; purely hypothetical technologies can ...
– new fields of technology, typically on the cutting edge. Examples include genetics, robotics, and nanotechnology (GNR). **
Artificial intelligence Artificial intelligence (AI) is intelligence—perceiving, synthesizing, and inferring information—demonstrated by machines, as opposed to intelligence displayed by animals and humans. Example tasks in which this is done include speech re ...
– computer mental abilities, especially those that previously belonged only to humans, such as
speech recognition Speech recognition is an interdisciplinary subfield of computer science and computational linguistics that develops methodologies and technologies that enable the recognition and translation of spoken language into text by computers with the m ...
,
natural language generation Natural language generation (NLG) is a software process that produces natural language output. In one of the most widely-cited survey of NLG methods, NLG is characterized as "the subfield of artificial intelligence and computational linguistics tha ...
, etc. ***
History of artificial intelligence The history of artificial intelligence (AI) began in antiquity, with myths, stories and rumors of artificial beings endowed with intelligence or consciousness by master craftsmen. The seeds of modern AI were planted by philosophers who attemp ...
(AI) *** Strong AI – hypothetical AI as smart as a human. **
Quantum computing Quantum computing is a type of computation whose operations can harness the phenomena of quantum mechanics, such as superposition, interference, and entanglement. Devices that perform quantum computations are known as quantum computers. Though ...
***
Timeline of quantum computing and communication This is a timeline of quantum computing. 1960s 1968 * Stephen Wiesner invents conjugate coding. (published in ACM SIGACT News 15(1):78–88) 1970s 1970 * James Park articulates the no-cloning theorem. 1973 * Alexander Holevo ...
*
Moore's law Moore's law is the observation that the number of transistors in a dense integrated circuit (IC) doubles about every two years. Moore's law is an observation and projection of a historical trend. Rather than a law of physics, it is an empir ...
– observation (not actually a
law Law is a set of rules that are created and are enforceable by social or governmental institutions to regulate behavior,Robertson, ''Crimes against humanity'', 90. with its precise definition a matter of longstanding debate. It has been vario ...
) that, over the
history of computing hardware The history of computing hardware covers the developments from early simple devices to aid calculation to modern day computers. Before the 20th century, most calculations were done by humans. The first aids to computation were purely mechan ...
, the number of
transistor upright=1.4, gate (G), body (B), source (S) and drain (D) terminals. The gate is separated from the body by an insulating layer (pink). A transistor is a semiconductor device used to Electronic amplifier, amplify or electronic switch, switch e ...
s on
integrated circuit An integrated circuit or monolithic integrated circuit (also referred to as an IC, a chip, or a microchip) is a set of electronic circuits on one small flat piece (or "chip") of semiconductor material, usually silicon. Large numbers of tiny ...
s doubles approximately every two years. The law is named after Intel co-founder
Gordon Moore Gordon Earle Moore (born January 3, 1929) is an American businessman, engineer, and the co-founder and chairman emeritus of Intel Corporation. He is also the original proponent of Moore's law. As of March 2021, Moore's net worth is rep ...
, who described the trend in his 1965 paper. *
Supercomputer A supercomputer is a computer with a high level of performance as compared to a general-purpose computer. The performance of a supercomputer is commonly measured in floating-point operations per second ( FLOPS) instead of million instructions ...
**
History of supercomputing The term supercomputing arose in the late 1920s in the United States in response to the IBM tabulators at Columbia University. The CDC 6600, released in 1964, is sometimes considered the first supercomputer. However, some earlier computers were c ...
*
Superintelligence A superintelligence is a hypothetical agent that possesses intelligence far surpassing that of the brightest and most gifted human minds. "Superintelligence" may also refer to a property of problem-solving systems (e.g., superintelligent language ...
*
Timeline of computing Timeline of computing presents events in the history of computing organized by year and grouped into six topic areas: predictions and concepts, first use and inventions, hardware systems and processors, operating systems, programming languages, an ...
*
Technological singularity The technological singularity—or simply the singularity—is a hypothetical future point in time at which technological growth becomes uncontrollable and irreversible, resulting in unforeseeable changes to human civilization. According to the m ...
– hypothetical point in the future when computer capacity rivals that of a human brain, enabling the development of strong AI — artificial intelligence at least as smart as a human. ** '' The Singularity Is Near'' – book by
Raymond Kurzweil Raymond Kurzweil ( ; born February 12, 1948) is an American computer scientist, author, inventor, and futurist. He is involved in fields such as optical character recognition (OCR), text-to-speech synthesis, speech recognition technology, and e ...
dealing with the progression and projections of development of computer capabilities, including beyond human levels of performance. *
TOP500 The TOP500 project ranks and details the 500 most powerful non-distributed computing, distributed computer systems in the world. The project was started in 1993 and publishes an updated list of the supercomputers twice a year. The first of these ...
– list of the 500 most powerful (non-distributed) computer systems in the world


References


External links


Historical and projected growth in supercomputer capacity
{{DEFAULTSORT:Orders Of Magnitude (Computing)
Computing Computing is any goal-oriented activity requiring, benefiting from, or creating computing machinery. It includes the study and experimentation of algorithmic processes, and development of both hardware and software. Computing has scientific, e ...
*